Traditional attention scales quadratically, making long-context processing prohibitively expensive.